首页> 外文OA文献 >Semantic Amodal Segmentation
【2h】

Semantic Amodal Segmentation

机译:语义模态分割

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Common visual recognition tasks such as classification, object detection, andsemantic segmentation are rapidly reaching maturity, and given the recent rateof progress, it is not unreasonable to conjecture that techniques for many ofthese problems will approach human levels of performance in the next few years.In this paper we look to the future: what is the next frontier in visualrecognition? We offer one possible answer to this question. We propose a detailed imageannotation that captures information beyond the visible pixels and requirescomplex reasoning about full scene structure. Specifically, we create an amodalsegmentation of each image: the full extent of each region is marked, not justthe visible pixels. Annotators outline and name all salient regions in theimage and specify a partial depth order. The result is a rich scene structure,including visible and occluded portions of each region, figure-ground edgeinformation, semantic labels, and object overlap. We create two datasets for semantic amodal segmentation. First, we label 500images in the BSDS dataset with multiple annotators per image, allowing us tostudy the statistics of human annotations. We show that the proposed full sceneannotation is surprisingly consistent between annotators, including for regionsand edges. Second, we annotate 5000 images from COCO. This larger datasetallows us to explore a number of algorithmic ideas for amodal segmentation anddepth ordering. We introduce novel metrics for these tasks, and along with ourstrong baselines, define concrete new challenges for the community.
机译:常见的视觉识别任务(例如分类,对象检测和语义分割)正在迅速达到成熟,并且鉴于最近的进展速度,可以推断出许多问题的技术在未来几年内将达到人类的表现水平,这并非没有道理。本文我们展望未来:视觉识别的下一个前沿是什么?我们为这个问题提供一个可能的答案。我们提出了一种详细的图像注释,它可以捕获可见像素以外的信息,并且需要对整个场景结构进行复杂的推理。具体来说,我们为每个图像创建一个无模态分段:标记每个区域的整个范围,而不仅仅是可见像素。注释器在图像中勾勒出轮廓并命名所有显着区域,并指定部分深度顺序。结果是丰富的场景结构,包括每个区域的可见和遮挡部分,图形底边缘信息,语义标签和对象重叠。我们为语义非模式分割创建了两个数据集。首先,我们在BSDS数据集中为500张图像标记每个图像带有多个注释器,从而使我们能够研究人类注释的统计信息。我们表明,拟议的完整场景批注在批注符之间(包括区域和边缘)是惊人地一致的。其次,我们注释来自COCO的5000张图像。更大的数据集使我们能够探索无模式分割和深度排序的许多算法思想。我们针对这些任务引入了新颖的指标,并结合了严格的基准,为社区定义了具体的新挑战。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号